Goto

Collaborating Authors

 critical system


Governable AI: Provable Safety Under Extreme Threat Models

Wang, Donglin, Liang, Weiyun, Chen, Chunyuan, Xu, Jing, Fu, Yulong

arXiv.org Artificial Intelligence

As AI rapidly advances, the security risks posed by AI are becoming increasingly severe, especially in critical scenarios, including those posing existential risks. If AI becomes uncontrollable, manipulated, or actively evades safety mechanisms, it could trigger systemic disasters. Existing AI safety approaches-such as model enhancement, value alignment, and human intervention-suffer from fundamental, in-principle limitations when facing AI with extreme motivations and unlimited intelligence, and cannot guarantee security. To address this challenge, we propose a Governable AI (GAI) framework that shifts from traditional internal constraints to externally enforced structural compliance based on cryptographic mechanisms that are computationally infeasible to break, even for future AI, under the defined threat model and well-established cryptographic assumptions.The GAI framework is composed of a simple yet reliable, fully deterministic, powerful, flexible, and general-purpose rule enforcement module (REM); governance rules; and a governable secure super-platform (GSSP) that offers end-to-end protection against compromise or subversion by AI. The decoupling of the governance rules and the technical platform further enables a feasible and generalizable technical pathway for the safety governance of AI. REM enforces the bottom line defined by governance rules, while GSSP ensures non-bypassability, tamper-resistance, and unforgeability to eliminate all identified attack vectors. This paper also presents a rigorous formal proof of the security properties of this mechanism and demonstrates its effectiveness through a prototype implementation evaluated in representative high-stakes scenarios.


Reinforcement Learning with Adaptive Regularization for Safe Control of Critical Systems

Neural Information Processing Systems

Reinforcement Learning (RL) is a powerful method for controlling dynamic systems, but its learning mechanism can lead to unpredictable actions that undermine the safety of critical systems. Here, we propose RL with Adaptive Regularization (RL-AR), an algorithm that enables safe RL exploration by combining the RL policy with a policy regularizer that hard-codes the safety constraints. RL-AR performs policy combination via a "focus module," which determines the appropriate combination depending on the state--relying more on the safe policy regularizer for less-exploited states while allowing unbiased convergence for well-exploited states. In a series of critical control applications, we demonstrate that RL-AR not only ensures safety during training but also achieves a return competitive with the standards of model-free RL that disregards safety.


Where AI Assurance Might Go Wrong: Initial lessons from engineering of critical systems

Bloomfield, Robin, Rushby, John

arXiv.org Artificial Intelligence

We draw on our experience working on system and software assurance and evaluation for systems important to society to summarise how safety engineering is performed in traditional critical systems, such as aircraft flight control. We analyse how this critical systems perspective might support the development and implementation of AI Safety Frameworks. We present the analysis in terms of: system engineering, safety and risk analysis, and decision analysis and support. We consider four key questions: What is the system? How good does it have to be? What is the impact of criticality on system development? and How much should we trust it? We identify topics worthy of further discussion. In particular, we are concerned that system boundaries are not broad enough, that the tolerability and nature of the risks are not sufficiently elaborated, and that the assurance methods lack theories that would allow behaviours to be adequately assured. We advocate the use of assurance cases based on Assurance 2.0 to support decision making in which the criticality of the decision as well as the criticality of the system are evaluated. We point out the orders of magnitude difference in confidence needed in critical rather than everyday systems and how everyday techniques do not scale in rigour. Finally we map our findings in detail to two of the questions posed by the FAISC organisers and we note that the engineering of critical systems has evolved through open and diverse discussion. We hope that topics identified here will support the post-FAISC dialogues.


What, Indeed, is an Achievable Provable Guarantee for Learning-Enabled Safety Critical Systems

Bensalem, Saddek, Cheng, Chih-Hong, Huang, Wei, Huang, Xiaowei, Wu, Changshun, Zhao, Xingyu

arXiv.org Artificial Intelligence

Machine learning has made remarkable advancements, but confidently utilising learning-enabled components in safety-critical domains still poses challenges. Among the challenges, it is known that a rigorous, yet practical, way of achieving safety guarantees is one of the most prominent. In this paper, we first discuss the engineering and research challenges associated with the design and verification of such systems. Then, based on the observation that existing works cannot actually achieve provable guarantees, we promote a two-step verification method for the ultimate achievement of provable statistical guarantees.


Study: Machine learning models cannot be trusted with absolute certainty

#artificialintelligence

An article titled "On misbehaviour and fault tolerance in machine learning systems," by doctoral researcher Lalli Myllyaho was named one of the best papers in 2022 by the Journal of Systems and Software. "The fundamental idea of the study is that if you put critical systems in the hands of artificial intelligence and algorithms, you should also learn to prepare for their failure," Myllyaho says. It may not necessarily be dangerous if a streaming service suggests uninteresting options to users, but such behavior undermines trust in the functionality of the system. However, faults in more critical systems that rely on machine learning can be much more harmful. "I wanted to investigate how to prepare for, for example, computer vision misidentifying things. For instance, in computed tomography artificial intelligence can identify objects in sections. If errors occur, it raises questions about to what extent computers should be trusted in such matters, and when to ask a human to take a look," says Myllyaho.


Empowering the trustworthiness of ML-based critical systems through engineering activities

Mattioli, Juliette, Delaborde, Agnes, Khalfaoui, Souhaiel, Lecue, Freddy, Sohier, Henri, Jurie, Frederic

arXiv.org Artificial Intelligence

This paper reviews the entire engineering process of trustworthy Machine Learning (ML) algorithms designed to equip critical systems with advanced analytics and decision functions. We start from the fundamental principles of ML and describe the core elements conditioning its trust, particularly through its design: namely domain specification, data engineering, design of the ML algorithms, their implementation, evaluation and deployment. The latter components are organized in an unique framework for the design of trusted ML systems.


Artificial Intelligence, Critical Systems, and the Control Problem - HS Today

#artificialintelligence

Artificial Intelligence (AI) is transforming our way of life from new forms of social organization and scientific discovery to defense and intelligence. This explosive progress is especially apparent in the subfield of machine learning (ML), where AI systems learn autonomously by identifying patterns in large volumes of data.[1] Indeed, over the last five years, the fields of AI and ML have witnessed stunning advancements in computer vision (e.g., object recognition), speech recognition, and scientific discovery.[2], Experts are increasingly voicing concerns over AI risk from misuse by state and non-state actors, principally in the areas of cybersecurity and disinformation propagation. However, issues of control – for example, how advanced AI decision-making aligns with human goals – are not as prominent in the discussion of risk and could ultimately be equally or more dangerous than threats from nefarious actors.


Toward Total-System Trustworthiness

Communications of the ACM

Communications' Inside Risks columns have long stressed the importance of total-system awareness of riskful situations, some of which may be very difficult to identify in advance. Specifically, the desired properties of the total system should be specified as requirements. Those desired properties are called emergent properties, because they often cannot be derived solely from lower-layer component properties, and appear only with respect to the total system. Unfortunately, additional behavior of the total system may arise--which either defeats the ability to satisfy the desired properties, or demonstrates that the set of required properties was improperly specified. In this column, I consider some cases in which total-system analysis is of vital importance, but generally very difficult to achieve with adequate assurance.


AI and Oncology: IRT Saint Exupéry and IUCT-Oncopole collaborate to advance cancer research - Actu IA

#artificialintelligence

On May 17, two Toulouse-based institutes, the IRT Saint Exupéry and the IUCT-Oncopole, a European center of expertise in oncology, signed a partnership focused on artificial intelligence. The aim of this partnership is to pool cutting-edge skills around AI-based research projects designed to improve prevention, diagnosis and care in oncology, particularly by predicting therapeutic effectiveness. Two of these projects are already at an advanced stage. The Saint Exupéry Institute of Technological Research aims to accelerate scientific and technological research and transfer to the aeronautics and space industries for the development of reliable, robust, certifiable and sustainable innovative solutions. A private research foundation supported by the French government, the IRT's mission is to promote French technological research for the benefit of industry and to develop the ecosystem of the aeronautics, space and critical systems sectors by providing access to its research projects, technological platforms and expertise.

  Country:
  Genre: Research Report (0.38)
  Industry: Health & Medicine > Therapeutic Area > Oncology (1.00)

Ensuring Dataset Quality for Machine Learning Certification

Picard, Sylvaine, Chapdelaine, Camille, Cappi, Cyril, Gardes, Laurent, Jenn, Eric, Lefèvre, Baptiste, Soumarmon, Thomas

arXiv.org Machine Learning

In this paper, we address the problem of dataset quality in the context of Machine Learning (ML)-based critical systems. We briefly analyse the applicability of some existing standards dealing with data and show that the specificities of the ML context are neither properly captured nor taken into ac-count. As a first answer to this concerning situation, we propose a dataset specification and verification process, and apply it on a signal recognition system from the railway domain. In addi-tion, we also give a list of recommendations for the collection and management of datasets. This work is one step towards the dataset engineering process that will be required for ML to be used on safety critical systems.